Efficient Orthogonal Parametrisation of Recurrent Neural Networks Using Householder Reflections
نویسندگان
چکیده
Recurrent Neural Networks (RNNs) have been successfully used in many applications. However, the problem of learning long-term dependencies in sequences using these networks is still a major challenge. Recent methods have been suggested to solve this problem by constraining the transition matrix to be unitary during training, which ensures that its norm is exactly equal to one. These methods either have limited expressiveness or scale poorly with the size of the network when compared with the simple RNN case, especially in an online learning setting. Our contributions are as follows. We first show that constraining the transition matrix to be unitary is a special case of an orthogonal constraint. Therefore, it may not be necessary to work with complex valued matrices. Then we present a new parametrisation of the transition matrix which allows efficient training of an RNN while ensuring that the matrix is always orthogonal. Using our approach, one online gradient step can, in the worst case, be performed in time complexity O(Tn), where T and n are the length of the input sequence and the size of the hidden layer respectively. This time complexity is the same as that of the simple RNN. Finally, we test our new †The Department of Computing and Information Systems. ‡Data61. parametrisation on problems with long-term dependencies. Our results suggest that the orthogonal constraint on the transition matrix has similar benefits to the unitary constraint.
منابع مشابه
Efficient Short-Term Electricity Load Forecasting Using Recurrent Neural Networks
Short term load forecasting (STLF) plays an important role in the economic and reliable operation ofpower systems. Electric load demand has a complex profile with many multivariable and nonlineardependencies. In this study, recurrent neural network (RNN) architecture is presented for STLF. Theproposed model is capable of forecasting next 24-hour load profile. The main feature in this networkis ...
متن کاملStabilizing Gradients for Deep Neural Networks via Efficient SVD Parameterization
Vanishing and exploding gradients are two of the main obstacles in training deep neural networks, especially in capturing long range dependencies in recurrent neural networks (RNNs). In this paper, we present an efficient parametrization of the transition matrix of an RNN that allows us to stabilize the gradients that arise in its training. Specifically, we parameterize the transition matrix by...
متن کاملOn an efficient implementation of the multistage Wiener filter through Householder reflections for DS-CDMA interference suppression
This paper describes the implementation of the multistage Wiener filter (MWF) through a series of nested Householder transformations applied to the input signal as a mean to form the analysis filter-bank part of the structure. The input signal is successively projected onto an appropriate subspace whereby the mutual information is maximized at each step, as required by the MWF structure. The me...
متن کاملRobust stability of stochastic fuzzy impulsive recurrent neural networks with\ time-varying delays
In this paper, global robust stability of stochastic impulsive recurrent neural networks with time-varyingdelays which are represented by the Takagi-Sugeno (T-S) fuzzy models is considered. A novel Linear Matrix Inequality (LMI)-based stability criterion is obtained by using Lyapunov functional theory to guarantee the asymptotic stability of uncertain fuzzy stochastic impulsive recurrent neural...
متن کاملOn Reflections and Rotations in Minkowski 3-space
Communicated by Abraham A. Ungar Abstract. In this paper, we investigate the reflections in Minkowski three-space by three different approaches. Firstly, we define Lorentzian reflections with Lorentzian inner product. Then, we examine Lorentzian reflections in view of Lorentzian Householder matrices. Finally, we use pure split quaternions to derive Lorentzian reflections. For each case, we find...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017